Goto

Collaborating Authors

 Wilmington


Tendon-Actuated Concentric Tube Endonasal Robot (TACTER)

Yamamoto, Kent K., Zachem, Tanner J., Kheradmand, Pejman, Zheng, Patrick, Abdelgadir, Jihad, Bailey, Jared Laurance, Pieter, Kaelyn, Codd, Patrick J., Chitalia, Yash

arXiv.org Artificial Intelligence

Endoscopic endonasal approaches (EEA) have become more prevalent for minimally invasive skull base and sinus surgeries. However, rigid scopes and tools significantly decrease the surgeon's ability to operate in tight anatomical spaces and avoid critical structures such as the internal carotid artery and cranial nerves. This paper proposes a novel tendon-actuated concentric tube endonasal robot (TACTER) design in which two tendon-actuated robots are concentric to each other, resulting in an outer and inner robot that can bend independently. The outer robot is a unidirectionally asymmetric notch (UAN) nickel-titanium robot, and the inner robot is a 3D-printed bidirectional robot, with a nickel-titanium bending member. In addition, the inner robot can translate axially within the outer robot, allowing the tool to traverse through structures while bending, thereby executing follow-the-leader motion. A Cosserat-rod based mechanical model is proposed that uses tendon tension of both tendon-actuated robots and the relative translation between the robots as inputs and predicts the TACTER tip position for varying input parameters. The model is validated with experiments, and a human cadaver experiment is presented to demonstrate maneuverability from the nostril to the sphenoid sinus. This work presents the first tendon-actuated concentric tube (TACT) dexterous robotic tool capable of performing follow-the-leader motion within natural nasal orifices to cover workspaces typically required for a successful EEA.


Box Pose and Shape Estimation and Domain Adaptation for Large-Scale Warehouse Automation

Yu, Xihang, Talak, Rajat, Shi, Jingnan, Viereck, Ulrich, Gilitschenski, Igor, Carlone, Luca

arXiv.org Artificial Intelligence

Modern warehouse automation systems rely on fleets of intelligent robots that generate vast amounts of data -- most of which remains unannotated. This paper develops a self-supervised domain adaptation pipeline that leverages real-world, unlabeled data to improve perception models without requiring manual annotations. Our work focuses specifically on estimating the pose and shape of boxes and presents a correct-and-certify pipeline for self-supervised box pose and shape estimation. We extensively evaluate our approach across a range of simulated and real industrial settings, including adaptation to a large-scale real-world dataset of 50,000 images. The self-supervised model significantly outperforms models trained solely in simulation and shows substantial improvements over a zero-shot 3D bounding box estimation baseline. Keywords: Certifiable models, computer vision, 3D robot vision, object pose estimation, safe perception, self-supervised learning.


The Radiance of Neural Fields: Democratizing Photorealistic and Dynamic Robotic Simulation

Nuthall, Georgina, Bowden, Richard, Mendez, Oscar

arXiv.org Artificial Intelligence

As robots increasingly coexist with humans, they must navigate complex, dynamic environments rich in visual information and implicit social dynamics, like when to yield or move through crowds. Addressing these challenges requires significant advances in vision-based sensing and a deeper understanding of socio-dynamic factors, particularly in tasks like navigation. To facilitate this, robotics researchers need advanced simulation platforms offering dynamic, photorealistic environments with realistic actors. Unfortunately, most existing simulators fall short, prioritizing geometric accuracy over visual fidelity, and employing unrealistic agents with fixed trajectories and low-quality visuals. To overcome these limitations, we developed a simulator that incorporates three essential elements: (1) photorealistic neural rendering of environments, (2) neurally animated human entities with behavior management, and (3) an ego-centric robotic agent providing multi-sensor output. By utilizing advanced neural rendering techniques in a dual-NeRF simulator, our system produces high-fidelity, photorealistic renderings of both environments and human entities. Additionally, it integrates a state-of-the-art Social Force Model to model dynamic human-human and human-robot interactions, creating the first photorealistic and accessible human-robot simulation system powered by neural rendering.


WALINET: A water and lipid identification convolutional Neural Network for nuisance signal removal in 1H MR Spectroscopic Imaging

Weiser, Paul, Langs, Georg, Motyka, Stanislav, Bogner, Wolfgang, Courvoisier, Sébastien, Hoffmann, Malte, Klauser, Antoine, Andronesi, Ovidiu C.

arXiv.org Artificial Intelligence

Purpose. Proton Magnetic Resonance Spectroscopic Imaging (1H-MRSI) provides non-invasive spectral-spatial mapping of metabolism. However, long-standing problems in whole-brain 1H-MRSI are spectral overlap of metabolite peaks with large lipid signal from scalp, and overwhelming water signal that distorts spectra. Fast and effective methods are needed for high-resolution 1H-MRSI to accurately remove lipid and water signals while preserving the metabolite signal. The potential of supervised neural networks for this task remains unexplored, despite their success for other MRSI processing. Methods. We introduce a deep-learning method based on a modified Y-NET network for water and lipid removal in whole-brain 1H-MRSI. The WALINET (WAter and LIpid neural NETwork) was compared to conventional methods such as the state-of-the-art lipid L2 regularization and Hankel-Lanczos singular value decomposition (HLSVD) water suppression. Methods were evaluated on simulated and in-vivo whole-brain MRSI using NMRSE, SNR, CRLB, and FWHM metrics. Results. WALINET is significantly faster and needs 8s for high-resolution whole-brain MRSI, compared to 42 minutes for conventional HLSVD+L2. Quantitative analysis shows WALINET has better performance than HLSVD+L2: 1) more lipid removal with 41% lower NRMSE, 2) better metabolite signal preservation with 71% lower NRMSE in simulated data, 155% higher SNR and 50% lower CRLB in in-vivo data. Metabolic maps obtained by WALINET in healthy subjects and patients show better gray/white-matter contrast with more visible structural details. Conclusions. WALINET has superior performance for nuisance signal removal and metabolite quantification on whole-brain 1H-MRSI compared to conventional state-of-the-art techniques. This represents a new application of deep-learning for MRSI processing, with potential for automated high-throughput workflow.


Novel Models for High-Dimensional Imaging: High-Resolution fMRI Acceleration and Quantification

Guo, Shouchang

arXiv.org Artificial Intelligence

The goals of functional Magnetic Resonance Imaging (fMRI) include high spatial and temporal resolutions with a high signal-to-noise ratio (SNR). To simultaneously improve spatial and temporal resolutions and maintain the high SNR advantage of OSSI, we present novel pipelines for fast acquisition and high-resolution fMRI reconstruction and physics parameter quantification. We propose a patch-tensor low-rank model, a physics-based manifold model, and a voxel-wise attention network. With novel models for acquisition and reconstruction, we demonstrate that we can improve SNR and resolution simultaneously without compromising scan time. All the proposed models outperform other comparison approaches with higher resolution and more functional information.


Liberty Defense Receives Letter of Intent from the Greater Toronto Airports Authority to acquire the HEXWAVE for use in Airport Security Programs.

#artificialintelligence

Toronto Pearson is located in Mississauga, west of Toronto, in Ontario, Canada. It is Canada's largest airport and the sixth-most-connected airport in the world. "As the first airport in the world to test HEXWAVE, we see the potential benefits of utilizing this innovative solution as part of our broader airport security program following further testing and evaluation," said Dwayne MacIntosh, Director, Corporate Safety and Security, GTAA. "We were impressed with the HEXWAVE's seamless screening during beta testing and look forward to working with Liberty Defense on the enhanced detection the HEXWAVE would bring to the airport." HEXWAVE uses millimeter wave, advanced 3D imaging, and AI to detect all types of concealed metallic and non-metallic weapons and other prohibited items – without having to divest common items.


3PL GEODIS deploying 1,000 more Locus Robotics AMRs

#artificialintelligence

Locus Robotics has signed what it claims to be one of the largest deployments of autonomous mobile robots (AMRs) ever. GEODIS is a leading global transport and logistics provider and has used Locus' AMRs since 2018, when it first deployed Locus' AMRs at a site in Indiana. The global third-party logistics company (3PL) has currently deployed Locus AMRs at 14 sites around the world, serving a variety of retail and consumer brands, including warehouses in the U.S and Europe. At press time, Locus hadn't provided how many of its AMRs GEODIS will have overall after these 1,000 are deployed. Locus told The Robot Report it doesn't have concrete evidence this is one of the largest AMR deals ever.


Validation and Generalizability of Self-Supervised Image Reconstruction Methods for Undersampled MRI

Yu, Thomas, Hilbert, Tom, Piredda, Gian Franco, Joseph, Arun, Bonanno, Gabriele, Zenkhri, Salim, Omoumi, Patrick, Cuadra, Meritxell Bach, Canales-Rodríguez, Erick Jorge, Kober, Tobias, Thiran, Jean-Philippe

arXiv.org Artificial Intelligence

Purpose: To investigate aspects of the validation of self-supervised algorithms for reconstruction of undersampled MR images: quantitative evaluation of prospective reconstructions, potential differences between prospective and retrospective reconstructions, suitability of commonly used quantitative metrics, and generalizability. Theory and Methods: Two self-supervised algorithms based on self-supervised denoising and neural network image priors were investigated. These methods are compared to a least squares fitting and a compressed sensing reconstruction using in-vivo and phantom data. Their generalizability was tested with prospectively under-sampled data from experimental conditions different to the training. Results: Prospective reconstructions can exhibit significant distortion relative to retrospective reconstructions/ground truth. Pixel-wise quantitative metrics may not capture differences in perceptual quality accurately, in contrast to a perceptual metric. All methods showed potential for generalization; generalizability is more affected by changes in anatomy/contrast than other changes. No-reference image metrics correspond well with human rating of image quality for studying generalizability. Compressed Sensing and learned denoising perform similarly well on all data. Conclusion: Self-supervised methods show promising results for accelerating image reconstruction in clinical routines. Nonetheless, more work is required to investigate standardized methods to validate reconstruction algorithms for future clinical use.


Warehouse Robotics Provider Berkshire Grey to Go Public Through SPAC Deal

WSJ.com: WSJD - Technology

Berkshire Grey shareholders Khosla Ventures, New Enterprise Associates, Canaan Partners and SoftBank Group Corp. will roll 100% of their equity into the combined company, the companies said Wednesday. Berkshire Grey, founded in 2013, develops systems that use artificial intelligence, mobile robots and scanning, gripping and sensing technology to pick orders and speed goods through distribution centers. The business had $35 million in revenue last year and expects to generate $59 million in revenue in 2021 and become profitable in 2024. Top news and in-depth analysis on the world of logistics, from supply chain to transport and technology. Its customers include Walmart Inc., Target Corp. and FedEx Corp.


Best Artificial Intelligence Logistics Startups -- Transmetrics Blog

#artificialintelligence

This article about the best artificial intelligence logistics startups is part of the "Logistics of the Future" series looking at the top logistics startups today. We are officially living in the age of Artificial Intelligence. It's everywhere we look, from AI-powered personal assistants to predictive analytics to making medical diagnoses, Artificial Intelligence is making incredible advances across all industries. In fact, a recent report on the state of Artificial Intelligence for enterprises found that supply chain and operations are some of the top areas where businesses are driving revenue from AI investment. Why is AI making such a big difference in the logistics and supply chain, particularly?